local identifiability
- Europe > France > Occitanie > Haute-Garonne > Toulouse (0.05)
- North America > United States > Arizona > Maricopa County > Scottsdale (0.04)
- Europe > Italy > Calabria > Catanzaro Province > Catanzaro (0.04)
- Asia > China (0.04)
Local Identifiability of Deep ReLU Neural Networks: the Theory
Is a sample rich enough to determine, at least locally, the parameters of a neural network? To answer this question, we introduce a new local parameterization of a given deep ReLU neural network by fixing the values of some of its weights. This allows us to define local lifting operators whose inverses are charts of a smooth manifold of a high dimensional space. The function implemented by the deep ReLU neural network composes the local lifting with a linear operator which depends on the sample. We derive from this convenient representation a geometrical necessary and sufficient condition of local identifiability. Looking at tangent spaces, the geometrical condition provides: 1/ a sharp and testable necessary condition of identifiability and 2/ a sharp and testable sufficient condition of local identifiability. The validity of the conditions can be tested numerically using backpropagation and matrix rank computations.
- North America > United States > California (0.14)
- North America > United States > Massachusetts > Hampshire County > Amherst (0.05)
- North America > Canada > British Columbia > Metro Vancouver Regional District > Vancouver (0.04)
- Europe > France > Occitanie > Haute-Garonne > Toulouse (0.05)
- North America > United States > Arizona > Maricopa County > Scottsdale (0.04)
- Europe > Italy > Calabria > Catanzaro Province > Catanzaro (0.04)
- Asia > China (0.04)
Review for NeurIPS paper: Improving Local Identifiability in Probabilistic Box Embeddings
Summary and Contributions: This paper is concerned with the mitigating local identifiability issues in probabilistic box embeddings. Probabilistic box embedding model is used to represent the probabilities of binary variables in terms of volumes of axis-aligned hyperrectangles. These probability distributions can be used to express specific relations between entities such as hierarchies, partial orders, and lattice structures. Learning box embeddings using gradient-based methods are not straightforward due to unidentifiability of such models. Non-identifiability means that the likelihood is not affected for whatever infinitesimal change in parameter space is made, hence relaxation of the boxes is required.
Local Identifiability of Deep ReLU Neural Networks: the Theory
Is a sample rich enough to determine, at least locally, the parameters of a neural network? To answer this question, we introduce a new local parameterization of a given deep ReLU neural network by fixing the values of some of its weights. This allows us to define local lifting operators whose inverses are charts of a smooth manifold of a high dimensional space. The function implemented by the deep ReLU neural network composes the local lifting with a linear operator which depends on the sample. We derive from this convenient representation a geometrical necessary and sufficient condition of local identifiability.
Local Identifiability of Deep ReLU Neural Networks: the Theory
Bona-Pellissier, Joachim, Malgouyres, François, Bachoc, François
Is a sample rich enough to determine, at least locally, the parameters of a neural network? To answer this question, we introduce a new local parameterization of a given deep ReLU neural network by fixing the values of some of its weights. This allows us to define local lifting operators whose inverses are charts of a smooth manifold of a high dimensional space. The function implemented by the deep ReLU neural network composes the local lifting with a linear operator which depends on the sample. We derive from this convenient representation a geometrical necessary and sufficient condition of local identifiability. Looking at tangent spaces, the geometrical condition provides: 1/ a sharp and testable necessary condition of identifiability and 2/ a sharp and testable sufficient condition of local identifiability. The validity of the conditions can be tested numerically using backpropagation and matrix rank computations.
- Europe > France > Occitanie > Haute-Garonne > Toulouse (0.04)
- North America > United States > Arizona > Maricopa County > Scottsdale (0.04)
- Europe > Italy > Calabria > Catanzaro Province > Catanzaro (0.04)
- Asia > China (0.04)
Improving Local Identifiability in Probabilistic Box Embeddings
Dasgupta, Shib Sankar, Boratko, Michael, Zhang, Dongxu, Vilnis, Luke, Li, Xiang Lorraine, McCallum, Andrew
Geometric embeddings have recently received attention for their natural ability to represent transitive asymmetric relations via containment. Box embeddings, where objects are represented by n-dimensional hyperrectangles, are a particularly promising example of such an embedding as they are closed under intersection and their volume can be calculated easily, allowing them to naturally represent calibrated probability distributions. The benefits of geometric embeddings also introduce a problem of local identifiability, however, where whole neighborhoods of parameters result in equivalent loss which impedes learning. Prior work addressed some of these issues by using an approximation to Gaussian convolution over the box parameters, however, this intersection operation also increases the sparsity of the gradient. In this work, we model the box parameters with min and max Gumbel distributions, which were chosen such that space is still closed under the operation of the intersection. The calculation of the expected intersection volume involves all parameters, and we demonstrate experimentally that this drastically improves the ability of such models to learn.
- North America > United States > California (0.14)
- North America > United States > Massachusetts > Hampshire County > Amherst (0.05)